PyDigger - unearthing stuff about Python


NameVersionSummarydate
phasellm 0.0.25 Wrappers for common large language models (LLMs) with support for evaluation. 2025-01-21 06:08:40
dyff 0.30.1 Meta-package to install the local SDK for the Dyff AI auditing platform. 2025-01-21 00:57:35
evalscope 0.10.0 EvalScope: Lightweight LLMs Evaluation Framework 2025-01-20 11:08:38
langsmith 0.2.11 Client library to connect to the LangSmith LLM Tracing and Evaluation Platform. 2025-01-17 00:25:37
vellum-uptrain-fork 0.7.2 Vellum UpTrain Fork 2025-01-16 22:17:04
evo 1.30.5 Python package for the evaluation of odometry and SLAM 2025-01-13 10:40:23
tokenization-scorer 1.1.8 Package for evaluating text tokenizations. 2025-01-13 10:36:40
sacrecomet 1.0.1 Tool to guide you through reporting the use of COMET for machine translation evaluation. 2025-01-13 10:19:24
arkas 0.0.1a15 Library to evaluate ML model performances 2025-01-12 20:54:10
agenta 0.31.0 The SDK for agenta is an open-source LLMOps platform. 2025-01-09 10:00:05
trajectopy 2.2.2 Trajectory Evaluation in Python 2025-01-09 08:34:22
latenpy 0.0.2 A package for lazy evaluation and caching to optimize scientific analysis workflows. 2025-01-05 17:52:29
trust_eval 0.1.0 Metric to measure RAG responses with inline citations 2025-01-03 04:05:05
mandoline 0.2.0 Official Python client for the Mandoline API 2025-01-01 02:43:28
opencompass 0.3.9 A comprehensive toolkit for large model evaluation 2024-12-31 09:29:20
dyff-schema 0.21.0 Data models for the Dyff AI auditing platform. 2024-12-27 15:28:27
frechet-music-distance 0.1.0 A library for computing Frechet Music Distance. 2024-12-24 17:46:33
zenoml-next 0.6.8 Upkeeping the now archived Zeno, the AI Data Management & Evaluation Platform 2024-12-21 21:36:55
modelradar 0.1.0 Aspect-based Forecasting Accuracy 2024-12-21 17:48:47
judges 0.0.5 A small library of research-backed LLM judges 2024-12-20 19:56:40
hourdayweektotal
5413987564281707
Elapsed time: 1.41862s